A Universal Approximation Theorem for Mixture-of-Experts Models
نویسندگان
چکیده
The mixture-of-experts (MoE) model is a popular neural network architecture for nonlinear regression and classification. The class of MoE mean functions is known to be uniformly convergent to any unknown target function, assuming that the target function is from a Sobolev space that is sufficiently differentiable and that the domain of estimation is a compact unit hypercube. We provide an alternative result, which shows that the class of MoE mean functions is dense in the class of all continuous functions over arbitrary compact domains of estimation. Our result can be viewed as a universal approximation theorem for MoE models. The theorem we present allows MoE users to be confident in applying such models for estimation when data arise from nonlinear and nondifferentiable generative processes.
منابع مشابه
Time Series Prediction using Mixtures of Experts
We consider the problem of prediction of stationary time series, using the architecture known as mixtures of experts (MEM). Here we suggest a mixture which blends several autoregressive models. This study focuses on some theoretical foundations of the prediction problem in this context. More precisely, it is demonstrated that this model is a universal approximator, with respect to learning the ...
متن کاملDiscontinuity-Sensitive Optimal Control Learning by Mixture of Experts
This paper proposes a discontinuity-sensitive approach to learn the solutions of parametric optimal control problems with high accuracy. Many tasks, ranging from model predictive control to reinforcement learning, may be solved by learning optimal solutions as a function of problem parameters. However, nonconvexity, discrete homotopy classes, and control switching cause discontinuity in the par...
متن کاملThe Basic Theorem and its Consequences
Let T be a compact Hausdorff topological space and let M denote an n–dimensional subspace of the space C(T ), the space of real–valued continuous functions on T and let the space be equipped with the uniform norm. Zukhovitskii [7] attributes the Basic Theorem to E.Ya.Remez and gives a proof by duality. He also gives a proof due to Shnirel’man, which uses Helly’s Theorem, now the paper obtains a...
متن کاملUniversal Approximation of Interval-valued Fuzzy Systems Based on Interval-valued Implications
It is firstly proved that the multi-input-single-output (MISO) fuzzy systems based on interval-valued $R$- and $S$-implications can approximate any continuous function defined on a compact set to arbitrary accuracy. A formula to compute the lower upper bounds on the number of interval-valued fuzzy sets needed to achieve a pre-specified approximation accuracy for an arbitrary multivariate con...
متن کاملUniversal Approximation by Ridge Computational Models and Neural Networks: A Survey
Computational models made up of linear combinations of ridge basis functions, widely used in machine learning and artificial intelligence, are considered. For such models, the literature on the so-called “universal approximation property” is surveyed. Different approaches, proof techniques, and tools are examined.
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- Neural computation
دوره 28 12 شماره
صفحات -
تاریخ انتشار 2016